FusionNet: Fusing via Fully-Aware Attention with Application to Machine Comprehension

نویسندگان

  • Hsin-Yuan Huang
  • Chenguang Zhu
  • Yelong Shen
  • Weizhu Chen
چکیده

This paper introduces a new neural structure called FusionNet, which extends existing attention approaches from three perspectives. First, it puts forward a novel concept of “history of word” to characterize attention information from the lowest word-level embedding up to the highest semantic-level representation. Second, it identifies an attention scoring function that better utilizes the “history of word” concept. Third, it proposes a fully-aware multi-level attention mechanism to capture the complete information in one text (such as a question) and exploit it in its counterpart (such as context or passage) layer by layer. We apply FusionNet to the Stanford Question Answering Dataset (SQuAD) and it achieves the first position for both single and ensemble model on the official SQuAD leaderboard at the time of writing (Oct. 4th, 2017). Meanwhile, we verify the generalization of FusionNet with two adversarial SQuAD datasets and it sets up the new state-of-the-art on both datasets: on AddSent, FusionNet increases the best F1 metric from 46.6% to 51.4%; on AddOneSent, FusionNet boosts the best F1 metric from 56.0% to 60.7%.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

FusionNet: A deep fully residual convolutional neural network for image segmentation in connectomics

Electron microscopic connectomics is an ambitious research direction with the goal of studying comprehensive brain connectivity maps by using high-throughput, nanoscale microscopy. One of the main challenges in connectomics research is developing scalable image analysis algorithms that require minimal user intervention. Recently, deep learning has drawn much attention in computer vision because...

متن کامل

Phase Conductor on Multi-layered Attentions for Machine Comprehension

Attention models have been intensively studied to improve NLP tasks such as machine comprehension via both question-aware passage attention model and selfmatching attention model. Our research proposes phase conductor (PhaseCond) for attention models in two meaningful ways. First, PhaseCond, an architecture of multi-layered attention models, consists of multiple phases each implementing a stack...

متن کامل

Bidirectional Attention Flow for Machine Comprehension

Machine comprehension (MC), answering a query about a given context paragraph, requires modeling complex interactions between the context and the query. Recently, attention mechanisms have been successfully extended to MC. Typically these methods use attention to focus on a small portion of the context and summarize it with a fixed-size vector, couple attentions temporally, and/or often form a ...

متن کامل

SocialFusion: Context-Aware Inference and Recommendation By Fusing Mobile, Sensor, and Social Data ; CU-CS-1059-09

Mobile social networks are rapidly becoming an important new domain showcasing the power of mobile computing systems. These networks combine mobile location information with social networking data to enable fully context-aware environments. This paper proposes SocialFusion, a framework to support context-aware inference and recommendation by fusing together mobile, sensor, and social data. We i...

متن کامل

SocialFusion: Context-Aware Inference and Recommendation By Fusing Mobile, Sensor, and Social Data

Mobile social networks are rapidly becoming an important new domain showcasing the power of mobile computing systems. These networks combine mobile location information with social networking data to enable fully context-aware environments. This paper proposes SocialFusion, a framework to support context-aware inference and recommendation by fusing together mobile, sensor, and social data. We i...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1711.07341  شماره 

صفحات  -

تاریخ انتشار 2017